Extremely Low Bit Neural Network: Squeeze the Last Bit Out with ADMM
نویسندگان
چکیده
Although deep learning models are highly effective for various tasks, such as detection and classification, the high computational cost prohibits the deployment in scenarios where either memory or computational resources are limited. In this paper, we focus on model compression and acceleration of deep models. We model a low bit quantized neural network as a constrained optimization problem. Then, with the use of ADMM, we decouple the discrete constraint and parameters of network. We also show how the resulting subproblems can be efficiently solved with extragradient and iterative quantization. The effectiveness of the proposed method has been demonstrated in extensive experiments on convolutional neural network for image recognition, object detection, and recurrent neural network for language model.
منابع مشابه
Optimum Drill Bit Selection by Using Bit Images and Mathematical Investigation
This study is designed to consider the two important yet often neglected factors, which are factory recommendation and bit features, in optimum bit selection. Image processing techniques have been used to consider the bit features. A mathematical equation, which is derived from a neural network model, is used for drill bit selection to obtain the bit’s maximum penetration rate that corresponds ...
متن کاملPredicting Force in Single Point Incremental Forming by Using Artificial Neural Network
In this study, an artificial neural network was used to predict the minimum force required to single point incremental forming (SPIF) of thin sheets of Aluminium AA3003-O and calamine brass Cu67Zn33 alloy. Accordingly, the parameters for processing, i.e., step depth, the feed rate of the tool, spindle speed, wall angle, thickness of metal sheets and type of material were selected as input and t...
متن کاملQuantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations
We introduce a method to train Quantized Neural Networks (QNNs) — neural networks with extremely low precision (e.g., 1-bit) weights and activations, at run-time. At traintime the quantized weights and activations are used for computing the parameter gradients. During the forward pass, QNNs drastically reduce memory size and accesses, and replace most arithmetic operations with bit-wise operati...
متن کاملModified 32-Bit Shift-Add Multiplier Design for Low Power Application
Multiplication is a basic operation in any signal processing application. Multiplication is the most important one among the four arithmetic operations like addition, subtraction, and division. Multipliers are usually hardware intensive, and the main parameters of concern are high speed, low cost, and less VLSI area. The propagation time and power consumption in the multiplier are always high. ...
متن کاملNew High Secure Network Steganography Method Based on Packet Length
In network steganography methods based on packet length, the length of the packets is used as a carrier for exchanging secret messages. Existing methods in this area are vulnerable against detections due to abnormal network traffic behaviors. The main goal of this paper is to propose a method which has great resistance to network traffic detections. In the first proposed method, the sender embe...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1707.09870 شماره
صفحات -
تاریخ انتشار 2017